Incorporating Gesture and Gaze into Multimodal Models of Human-to-Human Communication

نویسنده

  • Lei Chen
چکیده

In human communication, utterances are expressed with some structural events, e.g., sentence, speech repairs, control of floor, and etc. These structural events bring important information and quite helpful for a better understanding of human communication. Meanwhile, the human communication is also full of multimodal behaviors, e.g., gesture, gaze, and etc. As non-verbal signals, gesture and gaze show closed temporal and semantic links to spoken content. In my thesis, I am working on incorporating non-verbal cues in a multimodal model to better predict the structural events to further improve the understanding of human communication. Some research efforts made are summarized and my future research plan is described.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Human Language Technology Conference of the North American Chapter of the Association of Computational Linguistics Proceedings of the Doctoral Consortium

Structural information in language is important for obtaining a better understanding of a human communication (e.g., sentence segmentation, speaker turns, and topic segmentation). Human communication involves a variety of multimodal behaviors that signal both propositional content and structure, e.g., gesture, gaze, and body posture. These non-verbal signals have tight temporal and semantic lin...

متن کامل

Evaluating Factors Impacting the Accuracy of Forced Alignments in a Multimodal Corpus

People, when processing human-to-human communication, utilize everything they can in order to understand that communication, including speech and information such as the time and location of an interlocutor’s gesture and gaze. Speech and gesture are known to exhibit a synchronous relationship in human communication; however, the precise nature of that relationship requires further investigation...

متن کامل

A Gaze-Assisted Multimodal Approach to Rich and Accessible Human-Computer Interaction

Recent advancements in eye tracking technology are driving the adoption of gaze-assisted interaction as a rich and accessible human-computer interaction paradigm. Gaze-assisted interaction serves as a contextual, non-invasive, and explicit control method for users without disabilities; for users with motor or speech impairments, text entry by gaze serves as the primary means of communication. D...

متن کامل

Multimodal Dialogue for Ambient Intelligence and Smart Environments

Ambient Intelligence (AmI) and Smart Environments (SmE) are based on three foundations: ubiquitous computing, ubiquitous communication and intelligent adaptive interfaces [41]. This type of systems consists of a series of interconnected computing and sensing devices which surround the user pervasively in his environment and are invisible to him, providing a service that is dynamically adapted t...

متن کامل

The Workshop Programme Multimodal Corpora: Models Of Human Behaviour For The Specification And Evaluation Of Multimodal Input And Output Interfaces Tuesday 25th May 2004

We present our multimedia Visualization for Situated Temporal Analysis (VisSTA) system that facilitates analysis of multi-modal human communication incorporating video, audio, speech transcriptions, and coded multimodal (e.g. gesture and gaze) data. VisSTA is based on the Multiple Linked Representation strategy and keeps the user temporally situated by ensuring tight linkage among all represent...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2006